fix: remove seq_parallel + tp restriction in dtensor v2#1725
fix: remove seq_parallel + tp restriction in dtensor v2#1725
Conversation
Signed-off-by: Terry Kong <terryk@nvidia.com>
|
Warning Rate limit exceeded@terrykong has exceeded the limit for the number of commits that can be reviewed per hour. Please wait 2 minutes and 32 seconds before requesting another review. ⌛ How to resolve this issue?After the wait time has elapsed, a review can be triggered using the We recommend that you space out your commits to avoid hitting the rate limit. 🚦 How do rate limits work?CodeRabbit enforces hourly rate limits for each developer per organization. Our paid plans have higher rate limits than the trial, open-source and free plans. In all cases, we re-allow further reviews after a brief timeout. Please see our FAQ for further information. 📒 Files selected for processing (1)
Thanks for using CodeRabbit! It's free for OSS, and your support helps us grow. If you like it, consider giving us a shout-out. Comment |
|
| print( | ||
| "[WARNING]: sequence_parallel=True, but tp_size=1 which has no effect. Enable tp_size > 1 to use sequence parallelism." | ||
| ) | ||
| elif sequence_parallel_enabled and tp_size > 1: |
There was a problem hiding this comment.
shall we also remove this restriction in dtensor v1?
|
moving to draft from what @RayenTian found |
|
closing in favor of #1728 |
What does this PR do ?
Add a one line overview of what this PR aims to accomplish.
Issues
List issues that this PR closes (syntax):
Usage
# Add a code snippet demonstrating how to use thisBefore your PR is "Ready for review"
Pre checks:
Additional Information